76 research outputs found
Character-level Intra Attention Network for Natural Language Inference
Natural language inference (NLI) is a central problem in language
understanding. End-to-end artificial neural networks have reached
state-of-the-art performance in NLI field recently.
In this paper, we propose Character-level Intra Attention Network (CIAN) for
the NLI task. In our model, we use the character-level convolutional network to
replace the standard word embedding layer, and we use the intra attention to
capture the intra-sentence semantics. The proposed CIAN model provides improved
results based on a newly published MNLI corpus.Comment: EMNLP Workshop RepEval 2017: The Second Workshop on Evaluating Vector
Space Representations for NL
From Bilingual to Multilingual Neural Machine Translation by Incremental Training
Multilingual Neural Machine Translation approaches are based on the use of
task-specific models and the addition of one more language can only be done by
retraining the whole system. In this work, we propose a new training schedule
that allows the system to scale to more languages without modification of the
previous components based on joint training and language-independent
encoder/decoder modules allowing for zero-shot translation. This work in
progress shows close results to the state-of-the-art in the WMT task.Comment: Accepted paper at ACL 2019 Student Research Workshop. arXiv admin
note: substantial text overlap with arXiv:1905.0683
Chinese–Spanish neural machine translation enhanced with character and word bitmap fonts
Recently, machine translation systems based on neural networks have reached state-of-the-art results for some pairs of languages (e.g., German–English). In this paper, we are investigating the performance of neural machine translation in Chinese–Spanish, which is a challenging language pair. Given that the meaning of a Chinese word can be related to its graphical representation, this work aims to enhance neural machine translation by using as input a combination of: words or characters and their corresponding bitmap fonts. The fact of performing the interpretation of every word or character as a bitmap font generates more informed vectorial representations. Best results are obtained when using words plus their bitmap fonts obtaining an improvement (over a competitive neural MT baseline system) of almost six BLEU, five METEOR points and ranked coherently better in the human evaluation.Peer ReviewedPostprint (published version
- …